Papers
arxiv:2412.04862

EXAONE 3.5: Series of Large Language Models for Real-world Use Cases

Published on Dec 6, 2024
ยท Submitted by taesiri on Dec 9, 2024
#2 Paper of the day
Authors:
,
,
,
,
,
,
,
,
,

Abstract

EXAONE 3.5 language models, available in three configurations, demonstrate exceptional instruction following, long-context comprehension, and competitive performance across various benchmarks.

AI-generated summary

This technical report introduces the EXAONE 3.5 instruction-tuned language models, developed and released by LG AI Research. The EXAONE 3.5 language models are offered in three configurations: 32B, 7.8B, and 2.4B. These models feature several standout capabilities: 1) exceptional instruction following capabilities in real-world scenarios, achieving the highest scores across seven benchmarks, 2) outstanding long-context comprehension, attaining the top performance in four benchmarks, and 3) competitive results compared to state-of-the-art open models of similar sizes across nine general benchmarks. The EXAONE 3.5 language models are open to anyone for research purposes and can be downloaded from https://huggingface.co/LGAI-EXAONE. For commercial use, please reach out to the official contact point of LG AI Research: contact_us@lgresearch.ai.

Community

Paper submitter

EXAONE 3.5: Series of Large Language Models for Real-world Use Cases

Can this run on an Intel CPU? Can I run without Transformers?

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

The licensing makes the model solely academic, which is a pity.

Tell me in details about insurance of cancellation of holidays/ vacations. The main terms, words, word combinations and speech patterns, related to the subject

Sign up or log in to comment

Models citing this paper 17

Browse 17 models citing this paper

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2412.04862 in a dataset README.md to link it from this page.

Spaces citing this paper 11

Collections including this paper 8